Linearized two-layers neural networks in high dimension

نویسندگان

چکیده

We consider the problem of learning an unknown function f⋆ on d-dimensional sphere with respect to square loss, given i.i.d. samples {(yi,xi)}i≤n where xi is a feature vector uniformly distributed and yi=f⋆(xi)+εi. study two popular classes models that can be regarded as linearizations two-layers neural networks around random initialization: features model Rahimi–Recht (RF); tangent Jacot–Gabriel–Hongler (NT). Both these also randomized approximations kernel ridge regression (with different kernels), enjoy universal approximation properties when number neurons N diverges, for fixed dimension d. specific regimes: infinite-sample finite-width regime, in which n=∞ while d are large but finite, infinite-width finite-sample regime N=∞ n finite. In first we prove if dℓ+δ≤N≤dℓ+1−δ small δ>0, then RF effectively fits degree-ℓ polynomial raw features, NT degree-(ℓ+1) polynomial. second both reduce methods rotationally invariant kernels. that, sample size satisfies dℓ+δ≤n≤dℓ+1−δ, fit at most features. This lower bound achieved by regression, near-optimal prediction error vanishing regularization.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Multilayer Neural Networks: One or Two Hidden Layers?

We study the number of hidden layers required by a multilayer neu-ral network with threshold units to compute a function f from n d to {O, I}. In dimension d = 2, Gibson characterized the functions computable with just one hidden layer, under the assumption that there is no "multiple intersection point" and that f is only defined on a compact set. We consider the restriction of f to the neighbo...

متن کامل

Deep Neural Networks for High Dimension, Low Sample Size Data

Deep neural networks (DNN) have achieved breakthroughs in applications with large sample size. However, when facing high dimension, low sample size (HDLSS) data, such as the phenotype prediction problem using genetic data in bioinformatics, DNN suffers from overfitting and high-variance gradients. In this paper, we propose a DNN model tailored for the HDLSS data, named Deep Neural Pursuit (DNP)...

متن کامل

Multi-Layers Chaotic Neural Networks

Human brain can be regarded as a complicated multi-layers dynamical system which including plenty of chaotic neural groups. In this paper, a simplified multi-layers chaotic neural network is proposed and found that its dynamical characteristics are more complexity than those in a neural group through computer simulation.

متن کامل

Dimension Expansion of Neural Networks '

In this paper, we investigate the dimension expansion property of 3 layer feedforward neural networks and provide a helpful insight into how neural networks define complex decision boundaries. First, we note that adding a hidden neuron is equivalent to expanding the dimension of the space defined by the outputs of the hidden neurons. Thus, if the number of hidden neurons is larger than the numb...

متن کامل

VC Dimension of Neural Networks

This paper presents a brief introduction to Vapnik-Chervonenkis (VC) dimension, a quantity which characterizes the difficulty of distribution-independent learning. The paper establishes various elementary results, and discusses how to estimate the VC dimension in several examples of interest in neural network theory.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Annals of Statistics

سال: 2021

ISSN: ['0090-5364', '2168-8966']

DOI: https://doi.org/10.1214/20-aos1990